Variational Latent-State GPT for Semi-Supervised Task-Oriented Dialog Systems

نویسندگان

چکیده

Recently, two approaches, fine-tuning large pre-trained language models and variational training, have attracted significant interests, separately, for semi-supervised end-to-end task-oriented dialog (TOD) systems. In this paper, we propose Variational Latent-State GPT model (VLS-GPT), which is the first to combine strengths of approaches. Among many options models, generative inference learning TOD system, both as auto-regressive based on GPT-2, can be further trained over a mix labeled unlabeled data in manner. training VLS-GPT statistically computationally more challenging than previous works sequential latent variable use turn-level first-order Markovian. The non-Markovian due Transformer architecture. work, establish Recursive Monte Carlo Approximation (RMCA) objective with prove its unbiasedness. Further, develop computational strategy sampling-then-forward-computation realize RMCA, successfully overcomes memory explosion issue using speeds up training. Semi-supervised experiments are conducted benchmark multi-domain datasets different languages - MultiWOZ2.1 CrossWOZ. shown significantly outperform supervised-only self-training baselines.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Variational Autoencoder for Semi-Supervised Text Classification

Although semi-supervised variational autoencoder (SemiVAE) works in image classification task, it fails in text classification task if using vanilla LSTM as its decoder. From a perspective of reinforcement learning, it is verified that the decoder’s capability to distinguish between different categorical labels is essential. Therefore, Semi-supervised Sequential Variational Autoencoder (SSVAE) ...

متن کامل

A Semi Supervised Dialog Act Tagging for Telugu

In a task oriented domain, recognizing the intention of a speaker is important so that the conversation can proceed in the correct direction. This is possible only if there is a way of labeling the utterance with its proper intent. One such labeling techniques is Dialog Act (DA) tagging. This work focuses on discussing various n-gram DA tagging techniques. In this paper, a new method is propose...

متن کامل

Semi-Supervised Multi-Task Regression

Labeled data are needed for many machine learning applications but the amount available in some applications is scarce. Semi-supervised learning and multi-task learning are two of the approaches that have been proposed to alleviate this problem. In this paper, we seek to integrate these two approaches for regression applications. We first propose a new supervised multi-task regression method ca...

متن کامل

A Variational Approach to Semi-Supervised Clustering

We present a variational inference scheme for semi-supervised clustering in which data is supplemented with side information in the form of common labels. There is no mutual exclusion of classes assumption and samples are represented as a combinatorial mixture over multiple clusters. The method has other advantages such as the ability to find the most probable number of soft clusters in the dat...

متن کامل

Incremental dialog processing in a task-oriented dialog

Incremental Dialog Processing (IDP) enables Spoken Dialog Systems to gradually process minimal units of user speech in order to give the user an early system response. In this paper, we present an application of IDP that shows its effectiveness in a task-oriented dialog system. We have implemented an IDP strategy and deployed it for one month on a real-user system. We compared the resulting dia...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE/ACM transactions on audio, speech, and language processing

سال: 2023

ISSN: ['2329-9304', '2329-9290']

DOI: https://doi.org/10.1109/taslp.2023.3240661